634 research outputs found

    When Computer Vision Gazes at Cognition

    Get PDF
    Joint attention is a core, early-developing form of social interaction. It is based on our ability to discriminate the third party objects that other people are looking at. While it has been shown that people can accurately determine whether another person is looking directly at them versus away, little is known about human ability to discriminate a third person gaze directed towards objects that are further away, especially in unconstraint cases where the looker can move her head and eyes freely. In this paper we address this question by jointly exploring human psychophysics and a cognitively motivated computer vision model, which can detect the 3D direction of gaze from 2D face images. The synthesis of behavioral study and computer vision yields several interesting discoveries. (1) Human accuracy of discriminating targets 8{\deg}-10{\deg} of visual angle apart is around 40% in a free looking gaze task; (2) The ability to interpret gaze of different lookers vary dramatically; (3) This variance can be captured by the computational model; (4) Human outperforms the current model significantly. These results collectively show that the acuity of human joint attention is indeed highly impressive, given the computational challenge of the natural looking task. Moreover, the gap between human and model performance, as well as the variability of gaze interpretation across different lookers, require further understanding of the underlying mechanisms utilized by humans for this challenging task.Comment: Tao Gao and Daniel Harari contributed equally to this wor

    Do You See What I Mean? Visual Resolution of Linguistic Ambiguities

    Get PDF
    Understanding language goes hand in hand with the ability to integrate complex contextual information obtained via perception. In this work, we present a novel task for grounded language understanding: disambiguating a sentence given a visual scene which depicts one of the possible interpretations of that sentence. To this end, we introduce a new multimodal corpus containing ambiguous sentences, representing a wide range of syntactic, semantic and discourse ambiguities, coupled with videos that visualize the different interpretations for each sentence. We address this task by extending a vision model which determines if a sentence is depicted by a video. We demonstrate how such a model can be adjusted to recognize different interpretations of the same underlying sentence, allowing to disambiguate sentences in a unified fashion across the different ambiguity types.Comment: EMNLP 201

    Education Bill: Committee Stage Report. Research paper 11/37

    Get PDF
    "This is an account of the House of Commons Committee Stage of the Education Bill. It complements Research Paper 11/14, prepared for the Commons Second Reading debate... The Bill, as amended in Public Bill Committee, was published as Bill 180.

    Moth-inspired navigation algorithm in a turbulent odor plume from a pulsating source

    Full text link
    Some female moths attract male moths by emitting series of pulses of pheromone filaments propagating downwind. The turbulent nature of the wind creates a complex flow environment, and causes the filaments to propagate in the form of patches with varying concentration distributions. Inspired by moth navigation capabilities, we propose a navigation strategy that enables a flier to locate a pulsating odor source in a windy environment using a single threshold-based detection sensor. The strategy is constructed based on the physical properties of the turbulent flow carrying discrete puffs of odor and does not involve learning, memory, complex decision making or statistical methods. We suggest that in turbulent plumes from a pulsating point source, an instantaneously measurable quantity referred as a "puff crossing time", improves the success rate as compared to the navigation strategy based on "internal counter" that does not use this information. Using computer simulations of fliers navigating in turbulent plumes of the pulsating point source for varying flow parameters: turbulent intensities, plume meandering and wind gusts, we obtained trajectories qualitatively resembling male moths flights towards the pheromone sources. We quantified the probability of a successful navigation as well as the flight parameters such as the time spent searching and the total flight time, with respect to different turbulent intensities, meandering or gusts. The concepts learned using this model may help to design odor-based navigation of miniature airborne autonomous vehicles

    Budget 2016: a summary

    Get PDF
    • …
    corecore